What if you built the data backbone of an organization?
We are seeking a Senior Data Engineer based in Brussels, Belgium, to design, industrialize, and orchestrate data pipelines within a modern Lakehouse architecture
Your Responsibilities
- Design and maintain scalable data pipelines
- Develop ETL processes using SQL & Python
- Orchestrate data workflows with Airflow (or equivalent)
- Implement a Lakehouse architecture (Bronze / Silver / Gold layers)
- Model data (schemas, ERDs)
- Ensure data quality, reliability, and performance
- Apply Clean Code & CI/CD principles
- Containerize solutions using Docker
- Work with PySpark for distributed data processing
Ideal Profile
- 5+ years of experience in Data Engineering
- Proven experience with Lakehouse architecture
- Strong data modeling skills
- Autonomous, proactive, solution-oriented mindset
- English: minimum B2 level
- French or Dutch: C1 level
Solliciteren